Observable Changes of Hypotheses under Positive Reinforcement.

نویسندگان

  • P SUPPES
  • M SCHLAG-REY
چکیده

In mathematical models of concept learning it has consistently been assumed that positive reinforcement cannot lead to a change of the hypothesis determining the overt response. When hypotheses are experimentally identified and recorded along with positive and negative reinforcements of stimulus-response pairs, it can be shown that hypotheses may change after a positive reinforcement. Positive reinforcement has an information content for subjects that has not yet been adequately recognized in concept formation studies.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Reversibility of reinforcement between eating and running by schedule changes: A comparison of hypotheses and models

Rats increased eating that produced access to a running-wheel or increased running that produced access to food. depending on which response was potentially deprived. relative to baseline. by the scheduled ratio of responding. Under both schedules. instrumental responding significantly exceeded appropriate baselines of the noncontingent effects of the schedule. The results contradicted the hypo...

متن کامل

Monte Carlo Bayesian Reinforcement Learning

Bayesian reinforcement learning (BRL) encodes prior knowledge of the world in a model and represents uncertainty in model parameters by maintaining a probability distribution over them. This paper presents Monte Carlo BRL (MC-BRL), a simple and general approach to BRL. MC-BRL samples a priori a finite set of hypotheses for the model parameter values and forms a discrete partially observable Mar...

متن کامل

Reinforcement Learning for Problems with Hidden State

In this paper, we describe how techniques from reinforcement learning might be used to approach the problem of acting under uncertainty. We start by introducing the theory of partially observable Markov decision processes (POMDPs) to describe what we call hidden state problems. After a brief review of other POMDP solution techniques, we motivate reinforcement learning by considering an agent wi...

متن کامل

Reinforcement Learning for Decentralized Planning Under Uncertainty (Doctoral Consortium)

Decentralized partially-observable Markov decision processes (Dec-POMDPs) are a powerful tool for modeling multi-agent planning and decision-making under uncertainty. Prevalent Dec-POMDP solution techniques require centralized computation given full knowledge of the underlying model. But in real world scenarios, model parameters may not be known a priori, or may be difficult to specify. We prop...

متن کامل

Model-Based Probabilistic Pursuit via Inverse Reinforcement Learning

In this paper we address the integrated prediction, planning, and control problem that enables a single follower robot (the “photographer”) to quickly re-establish visual contact with a moving target (the “subject”) that has escaped the follower’s field of view. Our work addresses this unavoidable scenario, which reactive controllers are typically ill-equipped to handle, by making intelligent p...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • Science

دوره 148 3670  شماره 

صفحات  -

تاریخ انتشار 1965